Is the deconvolution layer the same as a convolutional layer?

نویسندگان

  • Wenzhe Shi
  • Jose Caballero
  • Lucas Theis
  • Ferenc Huszar
  • Andrew P. Aitken
  • Christian Ledig
  • Zehan Wang
چکیده

1 In our CVPR 2016 paper [1], we proposed a novel network architecture to perform single image super­resolution (SR). Most existing convolutional neural network (CNN) based super­resolution methods [10,11] first upsample the image using a bicubic interpolation, then apply a convolutional network. We will refer to these types of networks as high­resolution (HR) networks because the images are upsampled first. Instead, we feed the low­resolution (LR) input directly to a sub­pixel CNN as shown in Fig.1 : Figure 1: An illustration of the ESCPN framework where r denotes the upscaling ratio. Let denote the upscaling ratio ­ e.g if the input LR image is then the output HR image will be r 1 × 1. We then output number of channels instead of one high­resolution (HR) image and use periodic r × r r 2 shuffling to recreate the HR image. The exact details about how our efficient sub­pixel convolutional layer works can be found in the paper. We will refer to our network as a LR network. In this note, we want to focus on two aspects related to two questions most people asked us at CVPR when they saw this network. Firstly, how can channels magically become a HR image? And secondly, r 2 why are convolution in LR space a better choice? These are actually the key questions we tried to answer in the paper, but we were not able to go into as much depth and clarity as we would've liked given the page limit. To better answer these questions, we first discuss the relationships between the deconvolution layer in the form of the transposed convolution layer, the sub­pixel convolutional layer and our efficient sub­pixel convolutional layer, which we'll go through in Sec. 1 and Sec. 2. We will refer to our efficient sub­pixel convolutional layer as a convolutional layer in LR space to distinguish it from the common sub­pixel convolutional layer [5]. We will then show that for a fixed computational budget and complexity, a network with convolutions exclusively in LR space has more representation power at the same speed than a network that first upsamples the input in HR space.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Convolutional Neural Network based on Adaptive Pooling for Classification of Noisy Images

Convolutional neural network is one of the effective methods for classifying images that performs learning using convolutional, pooling and fully-connected layers. All kinds of noise disrupt the operation of this network. Noise images reduce classification accuracy and increase convolutional neural network training time. Noise is an unwanted signal that destroys the original signal. Noise chang...

متن کامل

Checkerboard artifact free sub-pixel convolution: A note on sub-pixel convolution, resize convolution and convolution resize

Convolutional neural networks (CNNs) are a popular and highly performant choice for pixel-wise dense prediction or generation. One of the commonly required components in such CNNs is a way to increase the resolution of the network’s input. The lower resolution inputs can be, for example, low-dimensional noise vectors in image generation [7] or low resolution (LR) feature maps for network vis...

متن کامل

An Information-Theoretic Discussion of Convolutional Bottleneck Features for Robust Speech Recognition

Convolutional Neural Networks (CNNs) have been shown their performance in speech recognition systems for extracting features, and also acoustic modeling. In addition, CNNs have been used for robust speech recognition and competitive results have been reported. Convolutive Bottleneck Network (CBN) is a kind of CNNs which has a bottleneck layer among its fully connected layers. The bottleneck fea...

متن کامل

Generative Deep Deconvolutional Learning

A generative model is developed for deep (multi-layered) convolutional dictionary learning. A novel probabilistic pooling operation is integrated into the deep model, yielding efficient bottom-up (pretraining) and top-down (refinement) probabilistic learning. After learning the deep convolutional dictionary, testing is implemented via deconvolutional inference. To speed up this inference, a new...

متن کامل

Compressive Sensing via Convolutional Factor Analysis

We solve the compressive sensing problem via convolutional factor analysis, where the convolutional dictionaries are learned in situ from the compressed measurements. An alternating direction method of multipliers (ADMM) paradigm for compressive sensing inversion based on convolutional factor analysis is developed. The proposed algorithm provides reconstructed images as well as features, which ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • CoRR

دوره abs/1609.07009  شماره 

صفحات  -

تاریخ انتشار 2016